Mml, Hybrid Bayesian Network Graphical Models, Statistical Consistency, Invariance and Uniqueness
نویسنده
چکیده
The problem of statistical — or inductive — inference pervades a large number of human activities and a large number of (human and non-human) actions requiring ‘intelligence’. Human and other ‘intelligent’ activity often entails making inductive inferences, remembering and recording observations from which one can make inductive inferences, learning (or being taught) the inductive inferences of others, and acting upon these inductive inferences. The Minimum Message Length (MML) approach to machine learning (within artificial intelligence) and statistical (or inductive) inference gives us a trade-off between simplicity of hypothesis (H) and goodness of fit to the data (D) [Wallace and Boulton, 1968, p. 185, sec 2; Boulton and Wallace, 1969; 1970, p. 64, col 1; Boulton, 1970; Boulton and Wallace, 1973b, sec. 1, col. 1; 1973c; 1975, sec 1 col 1; Wallace and Boulton, 1975, sec. 3; Boulton, 1975; Wallace and Georgeff, 1983; Wallace and Freeman, 1987; Wallace and Dowe, 1999a; Wallace, 2005; Comley and Dowe, 2005, secs. 11.1 and 11.4.1; Dowe, 2008a, sec 0.2.4, p. 535, col. 1 and elsewhere]. There are several different and intuitively appealing ways of thinking of MML. One such way is to note that files with structure compress (if our file compression program is able to find said structure) and that files without structure don’t compress. The more structure (that the compression program can find), the more the file will compress. Another, second, way to think of MML is in terms of Bayesian probability, where Pr(H) is the prior probability of a hypothesis, Pr(D|H) is the (statistical) likelihood of the data D given hypothesis H, − logPr(D|H) is the (negative) log-likelihood, Pr(H|D) is the posterior probability of H given D, and Pr(D) is the marginal probability of D — i.e., the probability that D will be generated (regardless of whatever the hypothesis might have been). Applying Bayes’s theorem twice, with or without the help of a Venn diagram, we have Pr(H|D) = Pr(H&D)/Pr(D) = (1/Pr(D)) Pr(H)Pr(D|H). Choosing the most probable hypothesis (a posteriori) is choosingH so as to maximise Pr(H|D). Given that Pr(D) and 1/Pr(D) are independent of the choice of
منابع مشابه
Learning Hybrid Bayesian Networks by MML
We use a Markov Chain Monte Carlo (MCMC) MML algorithm to learn hybrid Bayesian networks from observational data. Hybrid networks represent local structure, using conditional probability tables (CPT), logit models, decision trees or hybrid models, i.e., combinations of the three. We compare this method with alternative local structure learning algorithms using the MDL and BDe metrics. Results a...
متن کاملVisualizing Graphical Probabilistic Models
Complex probabilistic models are difficult to evaluate not only for consistency with the domain theory but also for novelty and significance. In this paper we describe the current visual representations of Bayesian networks and then present an implementation containing enhancements to these visualizations by integrating several well known techniques. We then apply our approach in an exploration...
متن کاملRule-based joint fuzzy and probabilistic networks
One of the important challenges in Graphical models is the problem of dealing with the uncertainties in the problem. Among graphical networks, fuzzy cognitive map is only capable of modeling fuzzy uncertainty and the Bayesian network is only capable of modeling probabilistic uncertainty. In many real issues, we are faced with both fuzzy and probabilistic uncertainties. In these cases, the propo...
متن کاملLearning with Bayesian Networks
A Bayesian network is a graphical tool which contains probabilistic relationships among various random variables. When used jointly with statistical techniques, the graphical model has several advantages for data analysis. First, a Bayesian network can be used to learn probabi l is t ic relationships, and thus can be used in a learning system to gain the understanding of a problem domain and to...
متن کاملPolyhedral Approach to Statistical Learning Graphical Models
The statistical task to learn graphical models of Bayesian network structure from data leads to the study of special polyhedra. In the paper, we offer an overview of our polyhedral approach to learning these statistical models. First, we report on the results on this topic from our recent papers. The second part of the paper brings some specific additional results inspired by this approach.
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2010